我们提出了一种用于语义分割的新型无监督域适应方法,该方法将训练的模型概括为源图像和相应的地面真相标签到目标域。域自适应语义分割的关键是学习域,不变和判别特征,而无需目标地面真相标签。为此,我们提出了一个双向像素 - 型对比型学习框架,该框架可最大程度地减少同一对象类特征的类内变化,同时无论域,无论域如何,都可以最大程度地提高不同阶层的阶层变化。具体而言,我们的框架将像素级特征与目标和源图像中同一对象类的原型保持一致(即分别为正面对),将它们设置为不同的类别(即负对),并执行对齐和分离在源图像中具有像素级特征的另一个方向的过程,目标图像中的原型。跨域匹配鼓励域不变特征表示,而双向像素 - 型对应对应关系汇总了同一对象类的特征,提供了歧视性特征。为了建立对比度学习的训练对,我们建议使用非参数标签转移(即跨不同域的像素 - 型对应关系,就可以生成目标图像的动态伪标签。我们还提出了一种校准方法,以补偿训练过程中逐渐补偿原型的阶级域偏差。
translated by 谷歌翻译
The Weather4Cast competition (hosted by NeurIPS 2022) required competitors to predict super-resolution rain movies in various regions of Europe when low-resolution satellite contexts covering wider regions are given. In this paper, we show that a general baseline 3D U-Net can be significantly improved with region-conditioned layers as well as orthogonality regularizations on 1x1x1 convolutional layers. Additionally, we facilitate the generalization with a bag of training strategies: mixup data augmentation, self-distillation, and feature-wise linear modulation (FiLM). Presented modifications outperform the baseline algorithms (3D U-Net) by up to 19.54% with less than 1% additional parameters, which won the 4th place in the core test leaderboard.
translated by 谷歌翻译
Applying suction grippers in unstructured environments is a challenging task because of depth and tilt errors in vision systems, requiring additional costs in elaborate sensing and control. To reduce additional costs, suction grippers with compliant bodies or mechanisms have been proposed; however, their bulkiness and limited allowable error hinder their use in complex environments with large errors. Here, we propose a compact suction gripper that can pick objects over a wide range of distances and tilt angles without elaborate sensing and control. The spring-inserted gripper body deploys and conforms to distant and tilted objects until the suction cup completely seals with the object and retracts immediately after, while holding the object. This seamless deployment and retraction is enabled by connecting the gripper body and suction cup to the same vacuum source, which couples the vacuum picking and retraction of the gripper body. Experimental results validated that the proposed gripper can pick objects within 79 mm, which is 1.4 times the initial length, and can pick objects with tilt angles up to 60{\deg}. The feasibility of the gripper was verified by demonstrations, including picking objects of different heights from the same picking height and the bin picking of transparent objects.
translated by 谷歌翻译
Recent advances in sophisticated synthetic speech generated from text-to-speech (TTS) or voice conversion (VC) systems cause threats to the existing automatic speaker verification (ASV) systems. Since such synthetic speech is generated from diverse algorithms, generalization ability with using limited training data is indispensable for a robust anti-spoofing system. In this work, we propose a transfer learning scheme based on the wav2vec 2.0 pretrained model with variational information bottleneck (VIB) for speech anti-spoofing task. Evaluation on the ASVspoof 2019 logical access (LA) database shows that our method improves the performance of distinguishing unseen spoofed and genuine speech, outperforming current state-of-the-art anti-spoofing systems. Furthermore, we show that the proposed system improves performance in low-resource and cross-dataset settings of anti-spoofing task significantly, demonstrating that our system is also robust in terms of data size and data distribution.
translated by 谷歌翻译
最近,变形金刚在空间范围内的学习和推断方面很受欢迎。但是,他们的性能依赖于存储并将注意力应用于视频中每个帧的功能张量。因此,随着视频的长度的增长,它们的空间和时间复杂性会线性增加,这对于长视频而言可能非常昂贵。我们提出了一种新颖的视觉记忆网络架构,用于空间范围的学习和推理问题。我们在内存网络中维护了固定的内存插槽,并提出了基于Gumbel-SoftMax的算法,以学习一种自适应策略以更新此内存。最后,该体系结构在视频对象细分(VOS)和视频预测问题上进行了基准测试。我们证明,我们的内存体系结构可实现最新的结果,在视频预测上优于基于变压器的方法和其他最新方法,同时保持恒定的内存能力与序列长度无关。
translated by 谷歌翻译